Mistral Small 3.1 24B Instruct 2503 GPTQ 4b 128g
Apache-2.0
This model is an INT4 quantized version of Mistral-Small-3.1-24B-Instruct-2503, using the GPTQ algorithm to reduce weights from 16-bit to 4-bit, significantly decreasing disk size and GPU memory requirements.